Goto

Collaborating Authors

 kronecker product







Approximating Real-Time Recurrent Learning with Random Kronecker Factors

Asier Mujika, Florian Meier, Angelika Steger

Neural Information Processing Systems

Wealso confirm these theoretical results experimentally. Further,we showempirically thattheKF-RTRLalgorithm captures long-term dependencies and almost matches the performance of TBPTT on real world tasks by trainingRecurrent Highway Networks on a synthetic string memorization task and onthe Penn TreeBank task, respectively.





2 Method Notations We use X>, X1, Tr(X) and vec(X) to denote the transpose, inverse, trace, and column-wise vectorization of amatrixX. We use X Y to represent the Kronecker product

Neural Information Processing Systems

In contrast, artificial agents are prone to'catastrophic forgetting' whereby performance on previous tasks deteriorates rapidly as new ones are acquired. This shortcoming has recently been addressed using methods that encourage parameters tostay close tothose used forprevious tasks.